Coexistence and local Mittag–Leffler stability of fractional-order recurrent neural networks with discontinuous activation functions
Huang Yu-Jiao, Chen Shi-Jun, Yang Xu-Hua, Xiao Jie
College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China

 

† Corresponding author. E-mail: hyj0507@zjut.edu.cn

Abstract
Abstract

In this paper, coexistence and local Mittag–Leffler stability of fractional-order recurrent neural networks with discontinuous activation functions are addressed. Because of the discontinuity of the activation function, Filippov solution of the neural network is defined. Based on Brouwerʼs fixed point theorem and definition of Mittag–Leffler stability, sufficient criteria are established to ensure the existence of equilibrium points, among which equilibrium points are locally Mittag–Leffler stable. Compared with the existing results, the derived results cover local Mittag–Leffler stability of both fractional-order and integral-order recurrent neural networks. Meanwhile discontinuous networks might have higher storage capacity than the continuous ones. Two numerical examples are elaborated to substantiate the effective of the theoretical results.

1. Introduction

Due to the wide applications in many fields, such as pattern recognition, associative memory, signal processing, optimization, and many other fields, artificial neural networks have been investigated in depth in recent decades.[15] Some applications, such as optimization, require network to have only one equilibrium point which is monostable. However, the multiple/stable equilibrium points of neural networks are demanded in some other applications such as associative memory and pattern recognition. The property that a neural network has multiple stable equilibrium points is called multistability of the neural network. When the neural network is applied for associative memory, the basic premise is to make sure that the neural network has a large number of stable equilibrium points. Therefore, the main purpose of investigating multistability of neural networks is to increase the number of stable equilibrium points.[6] Some considerable efforts have been devoted to investigating the issue. Multistability of two kinds of recurrent neural networks with activation functions symmetrical about the origin on the phase plane was investigated in [7]. Multiple μ-stability of neural networks with unbounded time-varying delays was considered in [8]. Multistability of recurrent neural networks with nonmonotonic activation functions and unbounded time-varying delays was discussed in [9]. Multistability in Mittag–Leffler sense of fractional-order neural networks with piecewise constant arguments was addressed in [10]. Many other works on multistability of recurrent neural networks can be found in [1115] and the references cited therein.

Fractional-order calculus which deals with derivatives and integrals in the form of noninteger order was established more than 300 years ago. Compared with the integer-order theory, the developments in the fractional-order theory have been rather slow. Lately, there has been a steady growth of interest in the fractional-order theory because lots of actual systems show fractional-order differential dynamical behaviors. A great number of applications about fractional-order derivative system have been found in science and engineering, such as drug absorption and disposition processes,[16] electromagnetic waves,[17] fractional damped oscillators and fractional forced oscillators,[18] viscoelastic systems,[19] heat conduction,[20] neural networks,[21] and so on. Noting that fractional calculus provides a powerful tool for describing memory and hereditary properties of systems.[22] So fractional-order derivative has been incorporated into neural networks. Meanwhile, fractional-order parameter can increase the flexibility of system parameter selection by introducing one more degree of freedom.[21] Fractional-order neural networks might be expected to play an important role in parameter estimation. Hence, the incorporation of memory term (a fractional derivative or integral) into neural network models is an extremely important improvement.[23] It is also necessary to study the fractional-order neural networks.

In recent years, dynamical behaviors of fractional-order recurrent neural networks have been investigated. α-stability and α-synchronization of fractional-order neural networks were presented in [24]. The authors discussed global Mittag–Leffler stability and synchronization of memristor-based fractional-order neural networks in [25]. The authors studied uniform stability of fractional-order complex-valued neural networks with time delays in [26]. The Mittag–Leffler stability and asymptotical ω-periodicity of fractional-order fuzzy neural networks were considered in [27]. The multiple Mittag–Leffler stability of fractional-order recurrent neural networks was investigated in [28]. An LMI-based uniform stability condition of fractional-order neural networks with time delay was established in [29]. Many other valuable works on dynamic analysis of fractional-order neural networks can be found in [23] and [3034] and the references cited therein.

It is well known that the activation functions play an important role in the dynamical analysis of recurrent neural networks. The storage capacity of pattern recognition and associative memory rely heavily on the structures of activation functions. As mentioned in [35], a brief overview on some common neural network models reveals that neural networks described by differential equations with a discontinuous right-hand side are of importance and do frequently arise in practice. When dealing with dynamical systems possessing high-slope nonlinear elements, it is often advantageous to model them with a system of differential equations with discontinuous right-hand side. Much efforts have been devoted to analyzing the dynamical properties of neural networks with discontinuous activation functions in [3638] and the references therein. There are also plenty of valuable works addressing the multistability issue of recurrent neural networks with discontinuous activation functions.[6,3941] However, to the best of our knowledge, most results concentrate upon the multistability of integral-order neural networks with discontinuous activation functions, and there are few results on the multistability of fractional-order ones with discontinuous activation functions. As a result, it is necessary to investigate the multistability of fractional-order neural networks with discontinuous activation functions.

Motivated by the above discussions, the purpose of this paper is to explore the coexistence and multiple Mittag–Leffler stability of fractional-order recurrent neural networks with discontinuous activation functions. The main contributions of this paper can be listed as follows:

(i) Based on the decomposition of state space, sufficient conditions are established for the existence of equilibrium points. Due to the incorporation of discontinuous activation functions, the neural networks might have higher storage capacity than the continuous ones.

(ii) By using the results in fractional-order derivative, sufficient conditions are established to ensure local Mittag–Leffler stability of equilibrium points. The speed of convergence in Mittag–Leffler sense is fast.

(iii) The results of fractional-order recurrent neural networks with discontinuous activation functions are extended to those of integral-order recurrent neural networks. Sufficient conditions are established to ensure coexistence and local exponential stability of integral-order recurrent neural networks.

The remainder parts of this paper consist of the following sections. Section 2 describes some preliminaries and problem formulations. The main results are stated in Section 3. Two examples are used to show the effectiveness of the obtained results in Section 4. Finally, the conclusion is drawn in Section 5.

Notations: The notation used in this paper is fairly standard. and denote, respectively, the left and the right limits of function at point . is the closure of the convex hull for set M. For a vector-valued function

define
where for and ‘×’ represents the Cartesian product.

2. Preliminaries and problem formulations

In this section, we present the neural network model discussed in this paper, and then give a brief introduction of the properties of fractional-order calculus, which includes the definitions and some useful lemmas.

Consider the following fractional-order recurrent neural networks:

where , α is the order of fractional-order differential equation (1) with represents the state of the i-th neuron at time t; is the self-regulating parameter of the neuron; aij is the connection weight of the j-th unit on the i-th unit; is the activation function; Ii represents the external input. We consider a general class of discontinuous activation function as follows:
where k is an integer that satisfies , are constants with , . Typical configuration of the activation function is depicted in Fig. 1.

Fig. 1. The graph for activation function (2).

Since is allowed to have points of discontinuity, we need to specify what mean by solution of the system having a discontinuous right-side. For this purpose, we consider the solution of fractional-order recurrent neural networks (1) in the sense of Filippov.[42]

Definition 1[36] For a given constant vector , the absolute continuous function x(t) with and is said to be a solution of system (1) on [0,T) if there exist measurable functions for almost all such that

where .

Definition 2[36] The point is said to be an equilibrium point of system (1) if

hold for all .

Definition 3[43] The fractional integral of order α for a function f is defined as

where and , is the Gamma function, that is .

Definition 4[43] Let f(t) is a differentiable function on , the Caputo fractional derivation of order α of f(t) is defined by

where , , and .

Definition 5[43] A two-parameter Mittag–Leffler function is defined by

where , , and . If , then . If , then .

Definition 6 Suppose is a positive invariant set, , is an equilibrium point of system (1). The equilibrium point is locally Mittag–Leffler stable, if for any , x(t) of system (1) satisfies

where , μ and σ are positive constants.

Lemma 1[28] Assume that , , and , is differentiable. If there exists such that and for , then .

Lemma 2[32] If , then for any

holds almost everywhere.

Lemma 3[25] If almost everywhere for , then for , where θ is a constant.

Lemma 4 (Brouwerʼs fixed point theory[44]) Suppose that M is a nonempty, convex, compact subset of , where , and that is a continuous mapping. Then f has a fixed point.

For the convenience of description, make some denotations as follows:

and denote
where or or . It is easy to see there exist subregions in . In the next section, we will investigate the existence and stability of the equilibrium point in every subregion of .

3. Main results

In this section, we will discuss the existence and stability for fractional-order recurrent neural networks with discontinuous activation functions. Some sufficient conditions will be established to ensure system (1) with activation function (2) can have equilibrium points, among which equilibrium points are locally Mittag–Leffler stable. Moreover, conditions for the existence and stability of equilibrium points for integral-order recurrent neural networks will be derived. The main results are stated one by one as follows. Firstly, we will discuss the existence of equilibrium points for system (1) with the characteristic that all the components of equilibrium points are in the continuous regions of activation function (2).

Theorem 1 There exist equilibrium points in for system (1) with activation function (2) if the following conditions hold for

Proof From conditions (3) and (4), there exists a small enough such that

Make some denotations as follows:

and denote
where or or , . It is easy to see there exist subregions in , and . Let be an arbitrary subregion located in . There exist the following three cases to discuss.

Case 1

When , , denote

It can be derived from condition (5) that

Case 2

When , , , denote

It can be derived from conditions (5) and (6) that

Case 3

When , , denote

It can be derived from condition (6) that

Denote , where are defined in expressions (7), (8), (9), respectively. Because is a continuous function in region , is a continuous function in region . Meanwhile is a nonempty, convex, compact subset of . is mapping the points in onto itself. Based on Brouwerʼs fixed point theorem, there exists a fixed point such that . It is easy to see that is the equilibrium point of system (1). As a result of arbitrariness of , there exists equilibrium points in for system (1) with activation function (2).

Next, we will discuss the existence of equilibrium points for system (1) with the characteristic that some components of equilibrium points are just the discontinuous points of activation function (2).

Theorem 2 If conditions (3) and (4) hold, system (1) with activation function (2) has equilibrium points with the characteristic that some components of them are just the discontinuous points of the activation function (2).

Proof Define , , . Let N(L) be the number of elements in L. Without loss of generality, suppose that , . Denote

where . Then there exist subregions in . From conditions (3) and (4), there exists a small enough such that

Make some denotations as follows:

and denote
where , . It is easy to see there exist subregions in , and . Let be an arbitrary subregion located in . There exist the following three cases to be discussed below.

Case 1

When , , denote

It can be derived from condition (10) that

Case 2

When , , , denote

It can be derived from conditions (10) and (11) that

Case 3

When , , denote

It can be derived from condition (11) that

Denote , where are defined in Eqs. (12), (13), (14), respectively. Because is a continuous function in region , is a continuous function in region . Meanwhile is a nonempty, convex, compact subset of . is mapping the points in onto itself. Based on Brouwerʼs fixed point theorem, then there exists one fixed point such that .

On the other hand, for , , we can gain from conditions (3) and (4) that

which infer that

Therefore, is an equilibrium point for system (1) with . By arbitrariness of , there are fixed points such that . Meanwhile, there are k+1 discontinuous points for activation function (2). So there are cases in which some components of are just the discontinuous points in activation function (2). Moreover, there are cases to choose components which are just the discontinuous points. It follows that there are equilibrium points for system (1) with activation function (2).

According to Theorem 1 and Theorem 2, we can obtain the following result.

Theorem 3 If the conditions (3) and (4) hold, there exist equilibrium points for the system (1) with activation function (2).

Proof Integrated Theorem 1 and Theorem 2, there exist equilibrium points for the system (1) with activation function (2). According to Newtonʼs binomial theorem, we can obtain that . This completes the proof.

Remark 1 Due to the incorporation of discontinuous activation functions, the neural networks might have higher storage capacity than the continuous ones.

In order to investigate the stability of equilibrium points of system (1) with activation function (2), it is necessary to discuss the invariance of the region .

Lemma 5 Suppose conditions (3) and (4) hold. The region is positively invariant.

Proof Let be an arbitrary subregion located in . We claim that the solution x(t) of system (1) will stay in for all . If this is not true, there exists some time such that escapes from ϕ. Without loss of generality, suppose that escapes from , , . Then it follows that

or
For the first case, we can obtain from Lemma 1. From conditions (3) and (4), it holds that

This yields a contradiction to . Similarly, the second case is either not true. Therefore, x(t) will stay in ϕ for all . That is, is positively invariant.

Next, we will investigate the stability of the equilibrium points in for system (1) with activation function (2).

Theorem 4 If conditions (3) and (4) hold, there exist local Mittag–Leffler stable equilibrium points in for system (1) with activation function (2).

Proof According to Theorem 1, there are equilibrium points for system (1) with activation function (2) in . Now we will discuss the stability of these equilibrium points. Let ϕ be an arbitrary subregion located in . is the equilibrium point located in ϕ.

Let , then

When , , . So,
According to Lemma 2
Choose some positive constants . Let , . Define
For any , there exists such that
Let , then we have
According to Lemma 1, , which implies that
Meanwhile we can obtain
where , . Hence,
According to Lemma 3
On the other hand
Obviously
Therefore, the equilibrium point located in ϕ is locally Mittag–Leffler stable. Due to the arbitrary of the picked subset ϕ, have regions like subset ϕ. Hence, there exist local Mittag–Leffler stable equilibrium points for system (1) with activation function (2).

Remark 2 Fractional-order neural networks have infinite memory and more degrees of freedom. As a result of these rewards, the fractional-order model is more accurate than the corresponding integer-order model, and they have more complicated properties. Moreover, the methods which are commonly utilized in multistability analysis of integer-order neural networks could not be directly extended to multistability analysis of fractional-order ones. To this end, Lemmas 1–3 are introduced in this paper.

Remark 3 Based on the configuration of activation function (2), the space can contain parts. Each of these parts is an attractive basin of one stable equilibrium point. If the length of any one section (for example ( ), ) in activation function (2) increases, then the attractive basin of the corresponding equilibrium point will be increased. We can gain that the attractive basins of equilibrium points rely heavily on the configuration of activation functions.

Remark 4 In [28], the multiple Mittag–Leffler stability of fractional-order recurrent neural networks has been investigated. The activation functions were required to be continuous, whereas for the discontinuous activation functions, the obtained results cannot be utilized.

It is worth noting that the above results can be generated in the following integral-order recurrent neural networks

The activation function is described as in activation function (2). According to the similar discussion of the above results, we can obtain the following corollaries.

Corollary 1 Suppose that conditions (3) and (4) hold. There exist equilibrium points in for system (16) with activation function (2).

Corollary 2 If conditions (3) and (4) hold, the system (16) with activation function (2) has equilibrium points with the characteristic that some components of them are just the discontinuous points of the activation function (2).

Corollary 3 If the conditions (3) and (4) hold, there exist equilibrium points for the system (1) with activation function (2).

Corollary 4 Suppose that conditions (3) and (4) hold. There exist local exponentially stable equilibrium points in for system (16) with activation function (2).

Remark 5 Actually, for integral-order recurrent neural networks (16), equation (15) can be rewritten as follows:

Hence, the equilibrium points of system (16) are locally exponentially stable.

Remark 6 According to Corollaries 1–4, we can gain that Theorem 4 includes the multistability of integer-order neural networks as a special case.

4. Illustrative examples

Example 1

Consider the following two-neuronal fractional-order recurrent neural network:

where α=0.9, , , , , . The activation function are defined as follows:
System (17) satisfies conditions (3) and (4). That is,
According to Theorems 1 and 4, the system (17) with activation function (18) has 9 local Mittag–Leffler stable equilibrium points. The dynamics of system (17) are illustrated in Figs. 24.

Fig. 2. Transient states of system (17).
Fig. 3. Dynamics of of system (17).
Fig. 4. Dynamics of of system (17).

According to Fig. 2, evolutions of 180 initial conditions have been tracked, which converge to 9 stable equilibrium points marked in red. The attractive basins of 9 stable equilibrium points are , , , , , , , , , respectively. Figure 3 and 4 also display initial conditions of x1 and x2 converge to 9 points. The effectiveness of the theoretical results has been verified.

Example 2

Consider the following 2-neuronal integral-order recurrent neural network:

Parameters are the same as those in Example 1. According to Corollaries 1 and 4, the system (19) with activation function (18) has 9 local exponentially stable equilibrium points. The dynamics of the system (19) are illustrated in Figs. 57. The effectiveness of the theoretical results has been verified.

Fig. 5. Transient states of system (19).
Fig. 6. Dynamics of of system (19).
Fig. 7. Dynamics of of system (19).

Remark 7 By calculating, 9 equilibrium points of fractional-order neural network (17) and integral-order neural network (19) are both as follows, (0.12,0.14), (0.3,1.04), (0.5,2.04), (1.02,0.5), (1.2,1.4), (1.4,2.4), (2.02,0.9), (2.2,1.8), (2.4,2.8). It can be seen that derivative of differential equations does not affect the distribution of equilibrium points. The derivative has a certain influence on the speed of convergence.

5. Conclusion

In this paper, multistability has been studied for fractional-order recurrent neural networks with a class of general discontinuous activation functions. It has shown that under some sufficient conditions, the system can have equilibrium points, of them are locally Mittag–Leffler stable. The results have been extended to the cases of integer-order recurrent neural networks. By introducing the discontinuous activation functions, the neural networks can have more equilibrium points in the whole state space. The theoretical results and simulations have demonstrated that discontinuous activation functions play an important role in increasing the quantity of equilibrium points of the system. The time delay of neural networks is unavoidable because of the finite speed of a signalʼs switch and transmission. There are few results about multistability of delayed fractional-order differential equations. So our further research work might be multistability of delayed fractional-order recurrent neural networks with discontinuous activation functions.

Reference
1 Chua L O Yang L 1988 IEEE Trans. Circ. Syst. 35 1273
2 Zhang H Wang Z Liu D 2008 IEEE Trans. Neural Netw. 19 855
3 Zhang H Wang Z Liu D 2014 IEEE Trans. Neural. Netw. Learn Syst. 25 1229
4 Bao G Zeng Z Shen Y 2018 Neural Netw. 98 51
5 Zheng Y J Zhou X H Sheng W G et al. 2018 Neural Netw. 102 78
6 Liang J Gong W Huang T 2016 Neural Netw. 84 125
7 Zeng Z Zheng W 2013 IEEE Trans. Neural. Netw. Learn Syst. 24 1749
8 Wang L Chen T 2014 Neural Netw. 53 109
9 Liu P Zeng Z Wang J 2018 IEEE Trans. Neural Netw. Learn. Syst. 29 3000
10 Wan L Wu A 2018 Neurocomputing 286 1
11 Cheng C Lin K Shih C 2007 Physica 225 61
12 Cao J Feng G Wang Y 2008 Physica 237 1734
13 Wang L Chen T 2012 IEEE Trans. Neural Netw. Learn. Syst. 23 1816
14 Huang Y Hu H 2015 Chin. Phys. 24 120701
15 Chen X Zhao Z Song Q et al. 2017 Appl. Math. Comput. 294 18
16 Dokoumetzidis A Macheras P 2009 J. Pharmacokinet Pharmacodyn 36 165
17 Heaviside O 1971 Electromagnetic Theory New York Chelsea
18 Zubair M Mughal M Naqvi Q 2011 Nonlin. Anal. Real World Appl. 12 2844
19 Koeller R 1984 J. Appl. Mech. 51 299
20 Jiang X Xu M 2010 Physica 389 3368
21 Chen L Chai Y Wu R et al. 2013 Neurocomputing 111 190
22 Kilbas A Srivastava H Trujillo J 2006 Theory and applications of fractional differential equations Boston Elsevier
23 Kaslik E Sivasundaram S 2012 Neural Netw. 32 245
24 Yu J Hu C Jiang H 2012 Neural Netw. 35 82
25 Chen J Zeng Z Jiang P 2014 Neural Netw. 51 1
26 Rakkiyappan R Cao J Velmurugan G 2015 IEEE Trans. Neural Netw. Learn. Syst. 26 84
27 Wu A Zeng Z 2016 Neural Netw. 74 73
28 Liu P Zeng Z Wang J 2017 IEEE Trans. Syst. Man. Cybern. Syst. 47 2279
29 Yang Y He Y Wang Y et al. 2018 Neurocomputing 285 82
30 Song C Cao J 2014 Neurocomputing 142 494
31 Stamova I 2014 Nonlinear Dyn. 77 1251
32 Liu H Li S Wang H 2017 Chin. Phys. 26 030504
33 Xi Y Yu Y Zhang S et al. 2018 Chin. Phys. 27 010202
34 Ding Z Zeng Z Wang L 2018 IEEE Trans. Neural Netw. Learn. Syst. 29 1477
35 Forti M Nistri P 2003 IEEE Trans. Circuits Syst. I. 50 1421
36 Forti M Nistri P Papini D 2005 IEEE Trans. Neural Netw. 16 1449
37 Li L Huang L 2009 Neurocomputing 72 3726
38 Wang D Huang L 2016 Nonlinear Dyn. 85 67
39 Bao G Zeng Z 2012 Neurocomputing 77 101
40 Huang Y Zhang H Wang Z 2012 Neurocomputing 91 21
41 Nie X Zheng W 2016 IEEE Trans. Cybern. 46 679
42 Filippov A 1988 Differential equations with discontinuous righthand sides, Mathematics and Its Applications (Soviet series) Boston Kluwer Academic
43 Podlubny I 1999 Fractional differential equations San Diego Academic Press
44 Brouwer L 1912 Mathmatische Annalen 71 97